mohit prabhushankar
- North America > United States > Georgia > Fulton County > Atlanta (0.04)
- Oceania > Australia > New South Wales > Sydney (0.04)
- Asia > Middle East > Jordan (0.04)
- North America > United States > Texas > Harris County > Houston (0.28)
- Oceania > New Zealand > North Island > Auckland Region > Auckland (0.04)
- North America > United States > Georgia > Fulton County > Atlanta (0.04)
- (3 more...)
- Research Report > Experimental Study (1.00)
- Research Report > New Finding (0.94)
- Research Report > Strength High (0.68)
- North America > United States > Georgia > Fulton County > Atlanta (0.04)
- Oceania > Australia > New South Wales > Sydney (0.04)
- Asia > Middle East > Jordan (0.04)
- North America > United States > Texas > Harris County > Houston (0.28)
- Oceania > New Zealand > North Island > Auckland Region > Auckland (0.04)
- North America > United States > Georgia > Fulton County > Atlanta (0.04)
- (3 more...)
- Research Report > Experimental Study (1.00)
- Research Report > New Finding (0.94)
- Research Report > Strength High (0.68)
CRACKS: Crowdsourcing Resources for Analysis and Categorization of Key Subsurface faults
Prabhushankar, Mohit, Kokilepersaud, Kiran, Quesada, Jorge, Yarici, Yavuz, Zhou, Chen, Alotaibi, Mohammad, AlRegib, Ghassan, Mustafa, Ahmad, Kumakov, Yusufjon
Crowdsourcing annotations has created a paradigm shift in the availability of labeled data for machine learning. Availability of large datasets has accelerated progress in common knowledge applications involving visual and language data. However, specialized applications that require expert labels lag in data availability. One such application is fault segmentation in subsurface imaging. Detecting, tracking, and analyzing faults has broad societal implications in predicting fluid flows, earthquakes, and storing excess atmospheric CO$_2$. However, delineating faults with current practices is a labor-intensive activity that requires precise analysis of subsurface imaging data by geophysicists. In this paper, we propose the $\texttt{CRACKS}$ dataset to detect and segment faults in subsurface images by utilizing crowdsourced resources. We leverage Amazon Mechanical Turk to obtain fault delineations from sections of the Netherlands North Sea subsurface images from (i) $26$ novices who have no exposure to subsurface data and were shown a video describing and labeling faults, (ii) $8$ practitioners who have previously interacted and worked on subsurface data, (iii) one geophysicist to label $7636$ faults in the region. Note that all novices, practitioners, and the expert segment faults on the same subsurface volume with disagreements between and among the novices and practitioners. Additionally, each fault annotation is equipped with the confidence level of the annotator. The paper provides benchmarks on detecting and segmenting the expert labels, given the novice and practitioner labels. Additional details along with the dataset links and codes are available at $\href{https://alregib.ece.gatech.edu/cracks-crowdsourcing-resources-for-analysis-and-categorization-of-key-subsurface-faults/}{link}$.
- Information Technology > Sensing and Signal Processing > Image Processing (1.00)
- Information Technology > Communications > Social Media > Crowdsourcing (1.00)
- Information Technology > Artificial Intelligence > Vision (1.00)
- Information Technology > Artificial Intelligence > Machine Learning > Neural Networks > Deep Learning (0.68)
Intelligent Multi-View Test Time Augmentation
Ozturk, Efe, Prabhushankar, Mohit, AlRegib, Ghassan
Personal use of this material is permitted. ABSTRACT In this study, we introduce an intelligent Test Time Augmentation (TTA) algorithm designed to enhance the robustness and accuracy of image classification models against viewpoint variations. Unlike traditional TTA methods that indiscriminately apply augmentations, our approach intelligently selects optimal augmentations based on predictive uncertainty Figure 1: Comparison of Intelligent Multi-View TTA with the metrics. This selection is achieved via a two-stage process: conventional single-view method. This illustrates how the intelligent the first stage identifies the optimal augmentation for each approach dynamically selects augmentation views to class by evaluating uncertainty levels, while the second stage refine predictions (P), in contrast to the conventional method's implements an uncertainty threshold to determine when applying reliance on a single, static view. This methodological advancement ensures that augmentations contribute to classification more effectively than a uniform application across the to the inference phase, applying augmentations to test data dataset.
- North America > United States > Georgia > Fulton County > Atlanta (0.04)
- Asia > Middle East > UAE > Abu Dhabi Emirate > Abu Dhabi (0.04)
Are Objective Explanatory Evaluation metrics Trustworthy? An Adversarial Analysis
Chowdhury, Prithwijit, Prabhushankar, Mohit, AlRegib, Ghassan, Deriche, Mohamed
Explainable AI (XAI) has revolutionized the field of deep learning by empowering users to have more trust in neural network models. The field of XAI allows users to probe the inner workings of these algorithms to elucidate their decision-making processes. The rise in popularity of XAI has led to the advent of different strategies to produce explanations, all of which only occasionally agree. Thus several objective evaluation metrics have been devised to decide which of these modules give the best explanation for specific scenarios. The goal of the paper is twofold: (i) we employ the notions of necessity and sufficiency from causal literature to come up with a novel explanatory technique called SHifted Adversaries using Pixel Elimination(SHAPE) which satisfies all the theoretical and mathematical criteria of being a valid explanation, (ii) we show that SHAPE is, infact, an adversarial explanation that fools causal metrics that are employed to measure the robustness and reliability of popular importance based visual XAI methods. Our analysis shows that SHAPE outperforms popular explanatory techniques like GradCAM and GradCAM++ in these tests and is comparable to RISE, raising questions about the sanity of these metrics and the need for human involvement for an overall better evaluation.
- Asia > Middle East > UAE > Abu Dhabi Emirate > Abu Dhabi (0.14)
- Asia > Middle East > UAE > Ajman Emirate > Ajman (0.04)
- North America > United States > Georgia > Fulton County > Atlanta (0.04)
- Europe > United Kingdom > England > Cambridgeshire > Cambridge (0.04)
Taxes Are All You Need: Integration of Taxonomical Hierarchy Relationships into the Contrastive Loss
Kokilepersaud, Kiran, Yarici, Yavuz, Prabhushankar, Mohit, AlRegib, Ghassan
In this work, we propose a novel supervised contrastive loss that enables the integration of taxonomic hierarchy information during the representation learning process. A supervised contrastive loss operates by enforcing that images with the same class label (positive samples) project closer to each other than images with differing class labels (negative samples). The advantage of this approach is that it directly penalizes the structure of the representation space itself. This enables greater flexibility with respect to encoding semantic concepts. However, the standard supervised contrastive loss only enforces semantic structure based on the downstream task (i.e. the class label). In reality, the class label is only one level of a \emph{hierarchy of different semantic relationships known as a taxonomy}. For example, the class label is oftentimes the species of an animal, but between different classes there are higher order relationships such as all animals with wings being ``birds". We show that by explicitly accounting for these relationships with a weighting penalty in the contrastive loss we can out-perform the supervised contrastive loss. Additionally, we demonstrate the adaptability of the notion of a taxonomy by integrating our loss into medical and noise-based settings that show performance improvements by as much as 7%.
- North America > United States > Georgia > Fulton County > Atlanta (0.04)
- Asia > Middle East > UAE > Abu Dhabi Emirate > Abu Dhabi (0.04)
- Information Technology > Artificial Intelligence > Representation & Reasoning (1.00)
- Information Technology > Artificial Intelligence > Machine Learning (1.00)
- Information Technology > Sensing and Signal Processing > Image Processing (0.94)
- Information Technology > Artificial Intelligence > Natural Language > Text Processing (0.90)
Counterfactual Gradients-based Quantification of Prediction Trust in Neural Networks
Prabhushankar, Mohit, AlRegib, Ghassan
The widespread adoption of deep neural networks in machine learning calls for an objective quantification of esoteric trust. In this paper we propose GradTrust, a classification trust measure for large-scale neural networks at inference. The proposed method utilizes variance of counterfactual gradients, i.e. the required changes in the network parameters if the label were different. We show that GradTrust is superior to existing techniques for detecting misprediction rates on $50000$ images from ImageNet validation dataset. Depending on the network, GradTrust detects images where either the ground truth is incorrect or ambiguous, or the classes are co-occurring. We extend GradTrust to Video Action Recognition on Kinetics-400 dataset. We showcase results on $14$ architectures pretrained on ImageNet and $5$ architectures pretrained on Kinetics-400. We observe the following: (i) simple methodologies like negative log likelihood and margin classifiers outperform state-of-the-art uncertainty and out-of-distribution detection techniques for misprediction rates, and (ii) the proposed GradTrust is in the Top-2 performing methods on $37$ of the considered $38$ experimental modalities. The code is available at: https://github.com/olivesgatech/GradTrust
- North America > United States > Georgia > Fulton County > Atlanta (0.04)
- North America > United States > California > Santa Clara County > San Jose (0.04)
Exploiting the Distortion-Semantic Interaction in Fisheye Data
Kokilepersaud, Kiran, Prabhushankar, Mohit, Yarici, Yavuz, AlRegib, Ghassan, Parchami, Armin
In this work, we present a methodology to shape a fisheye-specific representation space that reflects the interaction between distortion and semantic context present in this data modality. Fisheye data has the wider field of view advantage over other types of cameras, but this comes at the expense of high radial distortion. As a result, objects further from the center exhibit deformations that make it difficult for a model to identify their semantic context. While previous work has attempted architectural and training augmentation changes to alleviate this effect, no work has attempted to guide the model towards learning a representation space that reflects this interaction between distortion and semantic context inherent to fisheye data. We introduce an approach to exploit this relationship by first extracting distortion class labels based on an object's distance from the center of the image. We then shape a backbone's representation space with a weighted contrastive loss that constrains objects of the same semantic class and distortion class to be close to each other within a lower dimensional embedding space. This backbone trained with both semantic and distortion information is then fine-tuned within an object detection setting to empirically evaluate the quality of the learnt representation. We show this method leads to performance improvements by as much as 1.1% mean average precision over standard object detection strategies and.6% improvement over other state of the art representation learning approaches.
- North America > United States > Georgia > Fulton County > Atlanta (0.14)
- North America > United States > Michigan > Wayne County > Dearborn (0.04)
- Asia > Middle East > Republic of Türkiye (0.04)
- Automobiles & Trucks (0.68)
- Transportation (0.46)
- Health & Medicine (0.46)
- Education (0.46)
- Information Technology > Artificial Intelligence > Vision (1.00)
- Information Technology > Artificial Intelligence > Machine Learning > Statistical Learning (0.88)
- Information Technology > Artificial Intelligence > Machine Learning > Neural Networks > Deep Learning (0.68)
- Information Technology > Artificial Intelligence > Natural Language > Text Processing (0.67)